Goto

Collaborating Authors

 data scale


Supplementary Materials for On the Effects of Data Scale on Computer Control Agents

Neural Information Processing Systems

For completeness, in the following we include a datasheet based on the format of [1]. For what purpose was the dataset created? Was there a specific task in mind? Who created the dataset (e.g., which team, research group) and on behalf of which entity What do the instances that comprise the dataset represent (e.g., documents, photos, people, The dataset contains episodes of human demonstrations for mobile device control. How many instances are there in total (of each type, if appropriate)?






Scale-Aware Curriculum Learning for Ddata-Efficient Lung Nodule Detection with YOLOv11

Luo, Yi, Guo, Yike, Hooshangnejad, Hamed, Ding, Kai

arXiv.org Artificial Intelligence

Lung nodule detection in chest CT is crucial for early lung cancer diagnosis, yet existing deep learning approaches face challenges when deployed in clinical settings with limited annotated data. While curriculum learning has shown promise in improving model training, traditional static curriculum strategies fail in data-scarce scenarios. We propose Scale Adaptive Curriculum Learning (SACL), a novel training strategy that dynamically adjusts curriculum design based on available data scale. SACL introduces three key mechanisms:(1) adaptive epoch scheduling, (2) hard sample injection, and (3) scale-aware optimization. We evaluate SACL on the LUNA25 dataset using YOLOv11 as the base detector. Experimental results demonstrate that while SACL achieves comparable performance to static curriculum learning on the full dataset in mAP50, it shows significant advantages under data-limited conditions with 4.6%, 3.5%, and 2.0% improvements over baseline at 10%, 20%, and 50% of training data respectively. By enabling robust training across varying data scales without architectural modifications, SACL provides a practical solution for healthcare institutions to develop effective lung nodule detection systems despite limited annotation resources.


Demystifying Hybrid Thinking: Can LLMs Truly Switch Between Think and No-Think?

Wang, Shouren, Yang, Wang, Long, Xianxuan, Wang, Qifan, Chaudhary, Vipin, Han, Xiaotian

arXiv.org Artificial Intelligence

Hybrid thinking enables LLMs to switch between reasoning and direct answering, offering a balance between efficiency and reasoning capability. Y et our experiments reveal that current hybrid thinking LLMs only achieve partial mode separation: reasoning behaviors often leak into the no-think mode. To understand and mitigate this, we analyze the factors influencing controllability and identify four that matter most: (1) larger data scale, (2) using think and no-think answers from different questions rather than the same question, (3) a moderate increase in no-think data number, and (4) a two-phase strategy that first trains reasoning ability and then applies hybrid think training. Building on these findings, we propose a practical recipe that, compared to standard training, can maintain accuracy in both modes while significantly reducing no-think output length (from 1085 to 585 on MA TH500) and occurrences of reasoning-supportive tokens such as "wait" (from 5917 to 522 on MA TH500). Our findings highlight the limitations of current hybrid thinking and offer directions for strengthening its controllability. We compare the responses of Qwen3-8B under no-think and think modes. In the no-think mode, Qwen3-8B still performs reasoning outside the no-think constraint (e.g., generating reasoning-supportive words such as Wait), indicating that its hybrid thinking ability remains imperfect and cannot achieve full control.


Multimodal Language Models See Better When They Look Shallower

Chen, Haoran, Lin, Junyan, Chen, Xinghao, Fan, Yue, Dong, Jianfeng, Jin, Xin, Su, Hui, Fu, Jinlan, Shen, Xiaoyu

arXiv.org Artificial Intelligence

Multimodal large language models (MLLMs) typically extract visual features from the final layers of a pretrained Vision Transformer (ViT). This widespread deep-layer bias, however, is largely driven by empirical convention rather than principled analysis. While prior studies suggest that different ViT layers capture different types of information, with shallower layers focusing on fine visual details and deeper layers aligning more closely with textual semantics, the impact of this variation on MLLM performance remains underexplored. We present the first comprehensive study of visual layer selection for MLLMs, analyzing representation similarity across ViT layers to establish shallow, middle, and deep layer groupings. Through extensive evaluation of MLLMs (1.4B-7B parameters) across 10 benchmarks encompassing 60+ tasks, we find that while deep layers excel in semantic-rich tasks like OCR, shallow and middle layers significantly outperform them on fine-grained visual tasks including counting, positioning, and object localization. Building on these insights, we propose a lightweight feature fusion method that strategically incorporates shallower layers, achieving consistent improvements over both single-layer and specialized fusion baselines. Our work offers the first principled study of visual layer selection in MLLMs, showing that MLLMs can often see better when they look shallower.